Buchanan County
Appendix of ' Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation '
We calculate it for each sequence x and average over the whole corpus. When decoding auto-regressively, the probabilities of the repetitive sentence loops also have a self-reinforcement effect. As shown in Figure 2, the probability of the token'located' increases almost The work was conducted in Apple. Here we use the end token to split sentences for ease of experiments. We present the probability of the token'located' ( y-axis) as the number of historical repetitions Best viewed in color and zoomed in a desktop monitor.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > Missouri > Buchanan County > Saint Joseph (0.04)
- (6 more...)
- Media > Film (1.00)
- Government (1.00)
- Leisure & Entertainment > Sports > Basketball (0.46)
Appendix of ' Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation '
We calculate it for each sequence x and average over the whole corpus. When decoding auto-regressively, the probabilities of the repetitive sentence loops also have a self-reinforcement effect. As shown in Figure 2, the probability of the token'located' increases almost The work was conducted in Apple. Here we use the end token to split sentences for ease of experiments. We present the probability of the token'located' ( y-axis) as the number of historical repetitions Best viewed in color and zoomed in a desktop monitor.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > Missouri > Buchanan County > Saint Joseph (0.04)
- (6 more...)
- Media > Film (1.00)
- Government (1.00)
- Leisure & Entertainment > Sports > Basketball (0.46)
Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation
Xu, Jin, Liu, Xiaojiang, Yan, Jianhao, Cai, Deng, Li, Huayang, Li, Jian
While large-scale neural language models, such as GPT2 and BART, have achieved impressive results on various text generation tasks, they tend to get stuck in undesirable sentence-level loops with maximization-based decoding algorithms (\textit{e.g.}, greedy search). This phenomenon is counter-intuitive since there are few consecutive sentence-level repetitions in human corpora (e.g., 0.02\% in Wikitext-103). To investigate the underlying reasons for generating consecutive sentence-level repetitions, we study the relationship between the probabilities of the repetitive tokens and their previous repetitions in the context. Through our quantitative experiments, we find that 1) Language models have a preference to repeat the previous sentence; 2) The sentence-level repetitions have a \textit{self-reinforcement effect}: the more times a sentence is repeated in the context, the higher the probability of continuing to generate that sentence; 3) The sentences with higher initial probabilities usually have a stronger self-reinforcement effect. Motivated by our findings, we propose a simple and effective training method \textbf{DITTO} (Pseu\underline{D}o-Repet\underline{IT}ion Penaliza\underline{T}i\underline{O}n), where the model learns to penalize probabilities of sentence-level repetitions from pseudo repetitive data. Although our method is motivated by mitigating repetitions, experiments show that DITTO not only mitigates the repetition issue without sacrificing perplexity, but also achieves better generation quality. Extensive experiments on open-ended text generation (Wikitext-103) and text summarization (CNN/DailyMail) demonstrate the generality and effectiveness of our method.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > New York (0.04)
- (11 more...)
- Media > Film (1.00)
- Government (1.00)
- Law (0.86)
- (2 more...)
Efficient QUBO transformation for Higher Degree Pseudo Boolean Functions
Verma, Amit, Lewis, Mark, Kochenberger, Gary
Quadratic Unconstrained Binary Optimization (QUBO) is recognized as a unifying framework for modeling a wide range of problems. Problems can be solved with commercial solvers customized for solving QUBO and since QUBO have degree two, it is useful to have a method for transforming higher degree pseudo-Boolean problems to QUBO format. The standard transformation approach requires additional auxiliary variables supported by penalty terms for each higher degree term. This paper improves on the existing cubic-to-quadratic transformation approach by minimizing the number of additional variables as well as penalty coefficient. Extensive experimental testing on Max 3-SAT modeled as QUBO shows a near 100% reduction in the subproblem size used for minimization of the number of auxiliary variables.
- Oceania > New Zealand > North Island > Auckland Region > Auckland (0.04)
- North America > United States > Missouri > Buchanan County > Saint Joseph (0.04)
- North America > United States > Colorado > Denver County > Denver (0.04)
- (2 more...)
Goal Seeking Quadratic Unconstrained Binary Optimization
The Quadratic Unconstrained Binary Optimization (QUBO) modeling and solution framework is required for quantum and digital annealers whose goal is the optimization of a well defined metric, the objective function. However, diverse suboptimal solutions may be preferred over harder to implement strict optimal ones. In addition, the decision-maker usually has insights that are not always efficiently translated into the optimization model, such as acceptable target, interval or range values. Multi-criteria decision making is an example of involving the user in the decision process. In this paper, we present two variants of goal-seeking QUBO that minimize the deviation from the goal through a tabu-search based greedy one-flip heuristic. Experimental results illustrate the efficacy of the proposed approach over Constraint Programming for quickly finding a satisficing set of solutions.